Autonomous navigation at unsignalized intersections: A coupled reinforcement learning and model predictive control approach

نویسندگان

چکیده

This paper develops an integrated safety-enhanced reinforcement learning (RL) and model predictive control (MPC) framework for autonomous vehicles (AVs) to navigate unsignalized intersections. Researchers have extensively studied how AVs drive along highways. Nonetheless, intersections in urban environments remains a challenging task due the constant presence of moving road users, including turning vehicles, crossing or jaywalking pedestrians, cyclists. are thus required learn adapt dynamically evolving traffic environment. proposes design benchmark that allows sense real-time environment perform path planning. The agent generates curves feasible paths. ego vehicle attempts follow these paths under specific constraints. RL MPC navigation algorithms run parallel suitably selected enhance safety. AV is modeled with lateral longitudinal dynamics trained T-intersection using Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm various scenarios. It then tested on straight single multi-lane All experiments achieve desirable outcomes terms crash avoidance, driving efficiency, comfort, tracking accuracy. developed system provides adaptive can

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Autonomous HVAC Control, A Reinforcement Learning Approach

Recent high profile developments of autonomous learning thermostats by companies such as Nest Labs and Honeywell have brought to the fore the possibility of ever greater numbers of intelligent devices permeating our homes and working environments into the future. However, the specific learning approaches and methodologies utilised by these devices have never been made public. In fact little inf...

متن کامل

Autonomous vehicle navigation using evolutionary reinforcement learning

Reinforcement learning schemes perform direct on-line search in control space. This makes them appropriate for modifying control rules to obtain improvements in the performance of a system. The effectiveness of a reinforcement learning strategy is studied here through the training of a learning classz$er system (LCS) that controls the movement of an autonomous vehicle in simulated paths includi...

متن کامل

Autonomous UAV Navigation Using Reinforcement Learning

Unmanned aerial vehicles (UAV) are commonly used for missions in unknown environments, where an exact mathematical model of the environment may not be available. This paper provides a framework for using reinforcement learning to allow the UAV to navigate successfully in such environments. We conducted our simulation and real implementation to show how the UAVs can successfully learn to navigat...

متن کامل

Reinforcement Learning: Mdp Applied to Autonomous Navigation

The problem of autonomous vehicle navigation between lanes, around obstacles and towards a short term goal can be solved using Reinforcement Learning. The multi-lane road ahead of a vehicle may be represented by a Markov Decision Process (MDP) grid-world containing positive and negative rewards, allowing for practical computation of an optimal path using either value iteration (VI) or policy it...

متن کامل

Navigating Intersections with Autonomous Vehicles using Deep Reinforcement Learning

Providing an efficient strategy to navigate safely through unsignaled intersections is a difficult task that requires determining the intent of other drivers. We explore the effectiveness of Deep Reinforcement Learning to handle intersection problems. Using recent advances in Deep RL, we are able to learn policies that surpass the performance of a commonly-used heuristic approach in several met...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Transportation Research Part C-emerging Technologies

سال: 2022

ISSN: ['1879-2359', '0968-090X']

DOI: https://doi.org/10.1016/j.trc.2022.103662